ALGORITHM OF AMSGrad AND CHAOS OPTIMIZATION IN MULTILAYERED NEURON NETWORKS WITH STOCHASTIC GRADIENT DESCENT

نویسندگان

چکیده

In this paper, the AMSGrad stochastic optimization method was tested using logistic function that describes doubling process and Fourier spectra of error function. The implementation gradient descent algorithm, AMSGrad, done for a multilayer neural network with hidden layers. program recognizing printed digits written Python software environment. array each digit consisted set "0" "1" size 4x7. sample contained 5 possible distortions 3 arrays did not correspond to any digits. analysis influence value hyperparameters beta1 , beta2 learning rate on optimizing teaching network, which layers 28 neurons each, carried out. We constructed branching diagrams based these parameters. found hyperparameter contribution linear function, is associated number local global minima in retraining network. square contribution, block structure formation blocks processes. If alpha greater than rate, there transition chaotic state, accompanied by both multiple passage through minimum and, apparently, appearance minima. At such speed learning, optimizer practically does work, but presence i.e. gradient, general picture chaos described application shown lead better compared conventional even at optimal (the when existing doubles). Keywords: methods, diagrams.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Ant Colony Optimization and Stochastic Gradient Descent

In this article, we study the relationship between the two techniques known as ant colony optimization (ACO) and stochastic gradient descent. More precisely, we show that some empirical ACO algorithms approximate stochastic gradient descent in the space of pheromones, and we propose an implementation of stochastic gradient descent that belongs to the family of ACO algorithms. We then use this i...

متن کامل

Improved Stochastic gradient descent algorithm for SVM

In order to improve the efficiency and classification ability of Support vector machines (SVM) based on stochastic gradient descent algorithm, three algorithms of improved stochastic gradient descent (SGD) are used to solve support vector machine, which are Momentum, Nesterov accelerated gradient (NAG), RMSprop. The experimental results show that the algorithm based on RMSprop for solving the l...

متن کامل

Mutiple-gradient Descent Algorithm for Multiobjective Optimization

The steepest-descent method is a well-known and effective single-objective descent algorithm when the gradient of the objective function is known. Here, we propose a particular generalization of this method to multi-objective optimization by considering the concurrent minimization of n smooth criteria {J i } (i = 1,. .. , n). The novel algorithm is based on the following observation: consider a...

متن کامل

Distributed Stochastic Optimization via Adaptive Stochastic Gradient Descent

Stochastic convex optimization algorithms are the most popular way to train machine learning models on large-scale data. Scaling up the training process of these models is crucial in many applications, but the most popular algorithm, Stochastic Gradient Descent (SGD), is a serial algorithm that is surprisingly hard to parallelize. In this paper, we propose an efficient distributed stochastic op...

متن کامل

Stochastic Gradient Descent with GPGPU

We show how to optimize a Support Vector Machine and a predictor for Collaborative Filtering with Stochastic Gradient Descent on the GPU, achieving 1.66 to 6-times accelerations compared to a CPUbased implementation. The reference implementations are the Support Vector Machine by Bottou and the BRISMF predictor from the Netflix Prices winning team. Our main idea is to create a hash function of ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Elektronìka ta ìnformacìjnì tehnologìï

سال: 2023

ISSN: ['2224-0888', '2224-087X']

DOI: https://doi.org/10.30970/eli.21.7